@@ -74,40 +74,46 @@ module Agents |
||
74 | 74 |
|
75 | 75 |
def check |
76 | 76 |
Array(interpolated['url']).each do |url| |
77 |
- response = faraday.get(url) |
|
78 |
- if response.success? |
|
79 |
- feed = FeedNormalizer::FeedNormalizer.parse(response.body) |
|
80 |
- feed.clean! if interpolated['clean'] == 'true' |
|
81 |
- max_events = (interpolated['max_events_per_run'].presence || 0).to_i |
|
82 |
- created_event_count = 0 |
|
83 |
- feed.entries.sort_by { |entry| [entry.date_published, entry.last_updated] }.each.with_index do |entry, index| |
|
84 |
- break if max_events && max_events > 0 && index >= max_events |
|
85 |
- entry_id = get_entry_id(entry) |
|
86 |
- if check_and_track(entry_id) |
|
87 |
- created_event_count += 1 |
|
88 |
- create_event(payload: { |
|
89 |
- id: entry_id, |
|
90 |
- date_published: entry.date_published, |
|
91 |
- last_updated: entry.last_updated, |
|
92 |
- url: entry.url, |
|
93 |
- urls: entry.urls, |
|
94 |
- description: entry.description, |
|
95 |
- content: entry.content, |
|
96 |
- title: entry.title, |
|
97 |
- authors: entry.authors, |
|
98 |
- categories: entry.categories |
|
99 |
- }) |
|
100 |
- end |
|
101 |
- end |
|
102 |
- log "Fetched #{url} and created #{created_event_count} event(s)." |
|
103 |
- else |
|
104 |
- error "Failed to fetch #{url}: #{response.inspect}" |
|
105 |
- end |
|
77 |
+ check_url(url) |
|
106 | 78 |
end |
107 | 79 |
end |
108 | 80 |
|
109 | 81 |
protected |
110 | 82 |
|
83 |
+ def check_url(url) |
|
84 |
+ response = faraday.get(url) |
|
85 |
+ if response.success? |
|
86 |
+ feed = FeedNormalizer::FeedNormalizer.parse(response.body) |
|
87 |
+ feed.clean! if interpolated['clean'] == 'true' |
|
88 |
+ max_events = (interpolated['max_events_per_run'].presence || 0).to_i |
|
89 |
+ created_event_count = 0 |
|
90 |
+ feed.entries.sort_by { |entry| [entry.date_published, entry.last_updated] }.each.with_index do |entry, index| |
|
91 |
+ break if max_events && max_events > 0 && index >= max_events |
|
92 |
+ entry_id = get_entry_id(entry) |
|
93 |
+ if check_and_track(entry_id) |
|
94 |
+ created_event_count += 1 |
|
95 |
+ create_event(payload: { |
|
96 |
+ id: entry_id, |
|
97 |
+ date_published: entry.date_published, |
|
98 |
+ last_updated: entry.last_updated, |
|
99 |
+ url: entry.url, |
|
100 |
+ urls: entry.urls, |
|
101 |
+ description: entry.description, |
|
102 |
+ content: entry.content, |
|
103 |
+ title: entry.title, |
|
104 |
+ authors: entry.authors, |
|
105 |
+ categories: entry.categories |
|
106 |
+ }) |
|
107 |
+ end |
|
108 |
+ end |
|
109 |
+ log "Fetched #{url} and created #{created_event_count} event(s)." |
|
110 |
+ else |
|
111 |
+ error "Failed to fetch #{url}: #{response.inspect}" |
|
112 |
+ end |
|
113 |
+ rescue => e |
|
114 |
+ error "Failed to fetch #{url} with message '#{e.message}': #{e.backtrace}" |
|
115 |
+ end |
|
116 |
+ |
|
111 | 117 |
def get_entry_id(entry) |
112 | 118 |
entry.id.presence || Digest::MD5.hexdigest(entry.content) |
113 | 119 |
end |
@@ -133,4 +133,14 @@ describe Agents::RssAgent do |
||
133 | 133 |
expect(agent.memory['seen_ids']).to eq(agent.events.map {|e| Digest::MD5.hexdigest(e.payload['content']) }) |
134 | 134 |
end |
135 | 135 |
end |
136 |
+ |
|
137 |
+ describe 'logging errors with the feed url' do |
|
138 |
+ it 'includes the feed URL when an exception is raised' do |
|
139 |
+ mock(FeedNormalizer::FeedNormalizer).parse(anything) { raise StandardError.new("Some error!") } |
|
140 |
+ expect(lambda { |
|
141 |
+ agent.check |
|
142 |
+ }).not_to raise_error |
|
143 |
+ expect(agent.logs.last.message).to match(%r[Failed to fetch https://github.com]) |
|
144 |
+ end |
|
145 |
+ end |
|
136 | 146 |
end |